CokeBERT: Contextual knowledge selection and embedding towards enhanced pre-trained language models

نویسندگان

چکیده

Several recent efforts have been devoted to enhancing pre-trained language models (PLMs) by utilizing extra heterogeneous knowledge in graphs (KGs) and achieved consistent improvements on various knowledge-driven NLP tasks. However, most of these knowledge-enhanced PLMs embed static sub-graphs KGs ("knowledge context"), regardless that the required may change dynamically according specific text ("textual context"). In this paper, we propose a novel framework named Coke select contextual context textual for PLMs, which can avoid effect redundant ambiguous cannot match input text. Our experimental results show outperforms baselines typical tasks, indicating effectiveness dynamic understanding. Besides performance improvements, selected describe semantics text-related more interpretable form than conventional PLMs. source code datasets will be available provide details Coke.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Towards Enhanced Usability of Natural Language Interfaces to Knowledge Bases

One of the most prominent benefits gained from the emergence of Semantic Web technology is the possibility to access data more efficiently, through the use of ontologies [18]. Querying such data requires using formal languages such as SeRQL [7] or SPARQL [39]. However, the syntax of these formal languages tends to be too ’artificial’ and complex, especially for domain experts who are unfamiliar...

متن کامل

ImageNet pre-trained models with batch normalization

Convolutional neural networks (CNN) pre-trained on ImageNet are the backbone of most state-of-the-art approaches. In this paper, we present a new set of pretrained models with popular state-of-the-art architectures for the Caffe framework. The first release includes Residual Networks (ResNets) with generation script as well as the batch-normalization-variants of AlexNet and VGG19. All models ou...

متن کامل

Context Selection for Embedding Models

Word embeddings are an effective tool to analyze language. They have been recently extended to model other types of data beyond text, such as items in recommendation systems. Embedding models consider the probability of a target observation (a word or an item) conditioned on the elements in the context (other words or items). In this paper, we show that conditioning on all the elements in the c...

متن کامل

Functorial models for contextual pre-nets

The algebraic models of computation for contextual nets that have been proposed in the literature either rely on non-free monoid of objects, or introduce too many behaviors that must be somewhat filtered out. In this report, we exploit partial membership equational logic to define a suitable theory of models, where the meaningful concurrent computations can be selected by means of membership pr...

متن کامل

Contextual Information and Specific Language Models for Spoken Language Understanding

In this paper we explain how contextual expectations are generated and used in the task-oriented spoken language understanding system Dialogos. The hard task of recognizing spontaneous speech on the telephone may greatly benefit from the use of specific language models during the recognition of callers' utterances. By 'specific language models' we mean a set of language models that are trained ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: AI open

سال: 2021

ISSN: ['2666-6510']

DOI: https://doi.org/10.1016/j.aiopen.2021.06.004